Tightened L 0 - Relaxation Penalties for Classification 1 by Noam Goldberg , 2 Jonathan Eckstein
نویسندگان
چکیده
In optimization-based classification model selection, for example when using linear programming formulations, a standard approach is to penalize the L1 norm of some linear functional in order to select sparse models. Instead, we propose a novel integer linear program for sparse classifier selection, generalizing the minimum disagreement hyperplane problem whose complexity has been investigated in computational learning theory. Specifically, our mixed-integer problem is that of finding a separating hyperplane with minimum empirical error subject to an L0 penalty. We show that common “soft margin” linear programming formulations for robust classification are equivalent to a continuous relaxation of our model. Since the initial continuous relaxation is weak, we suggest a tighter relaxation, using novel cutting planes, to better approximate the integer solution. We describe a boosting algorithm, based on linear programming with dynamic generation of cuts and columns, that solves our relaxation. We demonstrate the classification performance of our proposed algorithm with experimental results, and justify our selection of parameters using a minimum description length, compression interpretation of learning.
منابع مشابه
Boosting Classifiers with Tightened L0-Relaxation Penalties
We propose a novel boosting algorithm which improves on current algorithms for weighted voting classification by striking a better balance between classification accuracy and the sparsity of the weight vector. In order to justify our optimization formulations, we first consider a novel integer linear program as a model for sparse classifier selection, generalizing the minimum disagreement halfs...
متن کاملSparse weighted voting classifier selection and its linear programming relaxations
Article history: Received 14 January 2011 Received in revised form 8 March 2012 Accepted 8 March 2012 Available online 9 March 2012 Communicated by W.-L. Hsu
متن کاملAn Improved Branch-and-Bound Method for Maximum Monomial Agreement
The NP-hard maximum monomial agreement (MMA) problem consists of finding a single logical conjunction that best fits a weighted dataset of “positive” and “negative” binary vectors. Computing classifiers using boosting methods involves a maximum agreement subproblem at each iteration, although such subproblems are typically solved by heuristic methods. Here, we describe an exact branch and bound...
متن کاملOptimization for Sparse and Accurate Classifiers
OF THE DISSERTATION Optimization for sparse and accurate classifiers by Noam Goldberg Dissertation Director: Professor Jonathan Eckstein Classification and supervised learning problems in general aim to choose a function that best describes a relation between a set of observed attributes and their corresponding outputs. We focus on binary classification, where the output is a binary response va...
متن کاملA New Perspective on Convex Relaxations of Sparse SVM
This paper proposes a convex relaxation of a sparse support vector machine (SVM) based on the perspective relaxation of mixed-integer nonlinear programs. We seek to minimize the zero-norm of the hyperplane normal vector with a standard SVM hinge-loss penalty and extend our approach to a zeroone loss penalty. The relaxation that we propose is a second-order cone formulation that can be efficient...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
دوره شماره
صفحات -
تاریخ انتشار 2009